Pairwise-Covariance Linear Discriminant Analysis

نویسندگان

  • Deguang Kong
  • Chris H. Q. Ding
چکیده

In machine learning, linear discriminant analysis (LDA) is a popular dimension reduction method. In this paper, we first provide a new perspective of LDA from an information theory perspective. From this new perspective, we propose a new formulation of LDA, which uses the pairwise averaged class covariance instead of the globally averaged class covariance used in standard LDA. This pairwise (averaged) covariance describes data distribution more accurately. The new perspective also provides a natural way to properly weigh different pairwise distances, which emphasizes the pairs of class with small distances, and this leads to the proposed pairwise covariance properly weighted LDA (pcLDA). The kernel version of pcLDA is presented to handle nonlinear projections. Efficient algorithms are presented to efficiently compute the proposed models. Introduction In the big data era, a large number of high-dimensional data (i.e., DNA microarray, social blog, image scenes, etc) are available for data analysis in different applications. Linear Discriminant Analysis (LDA) (Hastie, Tibshirani, and Friedman 2001) is one of the most popular methods for dimension reduction, which has shown state-of-the-art performance. The key idea of LDA is to find an optimal linear transformation which projects data into a low-dimensional space, where the data achieves maximum inter-class separability. The optimal solution to LDA is generally achieved by solving an eigenvalue problem. Despite the popularity and effectiveness of LDA, however, in standard LDA model, instead of emphasizing the pairwise-class distances, it simply takes an average of metrics computed in different pairs (i.e., computation of between-class scatter matrix Sb or within-class scatter matrix Sw). Thus, some pairwise class distances are depressed, especially for those pairs whose original class distances are relatively large. To overcome this issue, in this paper, we present a new formulation for pairwise linear discriminant analysis. To obtain a discriminant projection, the proposed method considers all the pairwise between-class and with-class distances. We call it “pairwise-covariance LDA (pcLDA)”. Then, the Copyright c 2014, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. pcLDA problem is cast into solving an optimization problem, which maximizes the class separability computed from pairwise distance. An efficient algorithm is proposed to solve the resultant problem, and experimental results indicate the good performance of the proposed method. A new perspective of LDA The standard linear discriminant analysis (LDA) is to seek a projection G = (g1, · · · ,gK 1) 2 <p⇥(K 1) which maximizes the class separability by solving,

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Weighted pairwise scatter to improve linear discriminant analysis

Linear Discriminant Analysis (LDA) aims to transform an original feature space to a lower dimensional space with as little loss in discrimination as possible. We introduce a novel LDA matrix computation that incorporates confusability information between classes into the transform. Our goal is to improve discrimination in LDA. In conventional LDA, a between class covariance matrix that is based...

متن کامل

Pareto models for discriminative multiclass linear dimensionality reduction

Linear Discriminant Analysis (LDA) is a popular tool for multiclass discriminative dimensionality reduction. However, LDA suffers from two major drawbacks: (i) it only optimizes the Bayes error for two-class problems where each class is a unimodal Gaussian with different mean, but both classes have equal full rank covariance matrices, and (ii) the multiclass extension does not maximize each pai...

متن کامل

Comparative Performance of Several Robust Linear Discriminant Analysis Methods

• The problem of the non-robustness of the classical estimates in the setting of the quadratic and linear discriminant analysis has been addressed by many authors: Todorov et al. [19, 20], Chork and Rousseeuw [1], Hawkins and McLachlan [4], He and Fung [5], Croux and Dehon [2], Hubert and Van Driessen [6]. To obtain high breakdown these methods are based on high breakdown point estimators of lo...

متن کامل

On Optimal Pairwise Linear Classi ers for Normal Distributions: The Two-Dimensional Case

Computing linear classi ers is a very important problem in statistical Pattern Recognition (PR). These classi ers have been investigated by the PR community extensively since they are the ones which are both easy to implement and comprehend. It is well known that when dealing with normally distributed classes, the optimal discriminant function for two-classes is linear only when the covariance ...

متن کامل

Fisher’s Linear Discriminant Analysis for Weather Data by reproducing kernel Hilbert spaces framework

Recently with science and technology development, data with functional nature are easy to collect. Hence, statistical analysis of such data is of great importance. Similar to multivariate analysis, linear combinations of random variables have a key role in functional analysis. The role of Theory of Reproducing Kernel Hilbert Spaces is very important in this content. In this paper we study a gen...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2014